Size matters: just how big is BIG?

نویسندگان

  • Paul R Burton
  • Anna L Hansell
  • Isabel Fortier
  • Teri A Manolio
  • Muin J Khoury
  • Julian Little
  • Paul Elliott
چکیده

BACKGROUND Despite earlier doubts, a string of recent successes indicates that if sample sizes are large enough, it is possible-both in theory and in practice-to identify and replicate genetic associations with common complex diseases. But human genome epidemiology is expensive and, from a strategic perspective, it is still unclear what 'large enough' really means. This question has critical implications for governments, funding agencies, bioscientists and the tax-paying public. Difficult strategic decisions with imposing price tags and important opportunity costs must be taken. METHODS Conventional power calculations for case-control studies disregard many basic elements of analytic complexity-e.g. errors in clinical assessment, and the impact of unmeasured aetiological determinants-and can seriously underestimate true sample size requirements. This article describes, and applies, a rigorous simulation-based approach to power calculation that deals more comprehensively with analytic complexity and has been implemented on the web as ESPRESSO: (www.p3gobservatory.org/powercalculator.htm). RESULTS Using this approach, the article explores the realistic power profile of stand-alone and nested case-control studies in a variety of settings and provides a robust quantitative foundation for determining the required sample size both of individual biobanks and of large disease-based consortia. Despite universal acknowledgment of the importance of large sample sizes, our results suggest that contemporary initiatives are still, at best, at the lower end of the range of desirable sample size. Insufficient power remains particularly problematic for studies exploring gene-gene or gene-environment interactions. Discussion Sample size calculation must be both accurate and realistic, and we must continue to strengthen national and international cooperation in the design, conduct, harmonization and integration of studies in human genome epidemiology.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Size matters: just how big is BIG? Quantifying realistic sample size requirements for human genome epidemiology

Background Despite earlier doubts, a string of recent successes indicates that if sample sizes are large enough, it is possible—both in theory and in practice—to identify and replicate genetic associations with common complex diseases. But human genome epidemiology is expensive and, from a strategic perspective, it is still unclear what ‘large enough’ really means. This question has critical im...

متن کامل

HYBRID ARTIFICIAL PHYSICS OPTIMIZATION AND BIG BANG-BIG CRUNCH ALGORITHM (HPBA) FOR SIZE OPTIMIZATION OF TRUSS STRUCTURES

Over the past decades, several techniques have been employed to improve the applicability of the metaheuristic optimization methods. One of the solutions for improving the capability of metaheuristic methods is the hybrid of algorithms. This study proposes a new optimization algorithm called HPBA which is based on the hybrid of two optimization algorithms; Big Bang-Big Crunch (BB-BC) inspired b...

متن کامل

Governments and Growth: Size Matters Country Size, Government Size, and Economic Growth

Big governments lead to lower levels of short-term economic growth. But is the size of this negative effect the same in different nations? I argue that big government matters less for growth in small countries than it does in large ones, because large countries are more dependent on domestic sources of production and consumption for their economic growth than are small countries. I find support...

متن کامل

Challenges and Opportunities with Big Data

The promise of data-driven decision-making is now being recognized broadly, and there is growing enthusiasm for the notion of “Big Data,” including the recent announcement from the White House about new funding initiatives across different agencies, that target research for Big Data. While the promise of Big Data is real – for example, it is estimated that Google alone contributed 54 billion do...

متن کامل

Lecture 2 — Algorithmic Cost Models

When we analyze the cost of an algorithm formally, we need to be reasonably precise in what model we are performing the analysis. Typically when analyzing algorithms the purpose of the model is not to calculate exact running times (this is too much to ask), but rather just to analyze asymptotic costs (i.e., big-O). These costs can then be used to compare algorithms in terms of how they scale to...

متن کامل

Lecture 2 — Algorithmic Cost Models Parallel and Sequential Data

When we analyze the cost of an algorithm formally, we need to be reasonably precise about the model we are performing the analysis in. Typically when analyzing algorithms the purpose of the model is not to calculate exact running times (this is too much to ask), but rather just to analyze asymptotic costs (i.e., big-O). These costs can then be used to compare algorithms in terms of how they sca...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • International Journal of Epidemiology

دوره 38  شماره 

صفحات  -

تاریخ انتشار 2009